Successive Rank-One Approximations of Nearly Orthogonally Decomposable Symmetric Tensors
نویسندگان
چکیده
Many idealized problems in signal processing, machine learning and statistics can be reduced to the problem of finding the symmetric canonical decomposition of an underlying symmetric and orthogonally decomposable (SOD) tensor. Drawing inspiration from the matrix case, the successive rank-one approximations (SROA) scheme has been proposed and shown to yield this tensor decomposition exactly, and a plethora of numerical methods have thus been developed for the tensor rank-one approximation problem. In practice, however, the inevitable errors (say) from estimation, computation, and modeling entail that the input tensor can only be assumed to be a nearly SOD tensor—i.e., a symmetric tensor slightly perturbed from the underlying SOD tensor. This article shows that even in the presence of perturbation, SROA can still robustly recover the symmetric canonical decomposition of the underlying tensor. It is shown that when the perturbation error is small enough, the approximation errors do not accumulate with the iteration number. Numerical results are presented to support the theoretical findings.
منابع مشابه
Successive Rank-One Approximations for Nearly Orthogonally Decomposable Symmetric Tensors
Many idealized problems in signal processing, machine learning and statistics can be reduced to the problem of finding the symmetric canonical decomposition of an underlying symmetric and orthogonally decomposable (SOD) tensor. Drawing inspiration from the matrix case, the successive rank-one approximations (SROA) scheme has been proposed and shown to yield this tensor decomposition exactly, an...
متن کاملOrthogonal Tensor Decomposition
In symmetric tensor decomposition one expresses a given symmetric tensor T a sum of tensor powers of a number of vectors: T = v⊗d 1 + · · · + v ⊗d k . Orthogonal decomposition is a special type of symmetric tensor decomposition in which in addition the vectors v1, ..., vk are required to be pairwise orthogonal. We study the properties of orthogonally decomposable tensors. In particular, we give...
متن کاملOn the Tensor Svd and Optimal Low Rank Orthogonal Approximations of Tensors
Abstract. It is known that a high order tensor does not necessarily have an optimal low rank approximation, and that a tensor might not be orthogonally decomposable (i.e., admit a tensor SVD). We provide several sufficient conditions which lead to the failure of the tensor SVD, and characterize the existence of the tensor SVD with respect to the Higher Order SVD (HOSVD) of a tensor. In face of ...
متن کاملVanishing of Doubly Symmetrized Tensors
Symmetrizations of tensors by irreducible characters of the symmetric group serve as natural analogues of symmetric and skew-symmetric tensors. The question of when a symmetrized decomposable tensor is non-zero is intimately related to the rank partition of a matroid extracted from the tensor. In this paper we characterize the non-vanishing of the symmetrization of certain partially symmetrized...
متن کاملTensor Decompositions via Two-Mode Higher-Order SVD (HOSVD)
Tensor decompositions have rich applications in statistics and machine learning, and developing efficient, accurate algorithms for the problem has received much attention recently. Here, we present a new method built on Kruskal’s uniqueness theorem to decompose symmetric, nearly orthogonally decomposable tensors. Unlike the classical higher-order singular value decomposition which unfolds a ten...
متن کامل